Word Acquisition in Neural Language Models
نویسندگان
چکیده
Abstract We investigate how neural language models acquire individual words during training, extracting learning curves and ages of acquisition for over 600 on the MacArthur-Bates Communicative Development Inventory (Fenson et al., 2007). Drawing studies word in children, we evaluate multiple predictors words’ LSTMs, BERT, GPT-2. find that effects concreteness, length, lexical class are pointedly different children models, reinforcing importance interaction sensorimotor experience child acquisition. Language rely far more frequency than but, like they exhibit slower longer utterances. Interestingly, follow consistent patterns training both unidirectional bidirectional LSTM Transformer architectures. Models predict based unigram token frequencies early before transitioning loosely to bigram probabilities, eventually converging nuanced predictions. These results shed light role distributional mechanisms while also providing insights human-like models.
منابع مشابه
Word Order Acquisition in Persian Speaking Children
Objectives: Persian is a pro-drop language with canonical Subject-Object-Verb (SOV) word order. This study investigates the acquisition of word order in Persian-speaking children. Methods: In the present study, participants were 60 Persian-speaking children (30 girls and 30 boys) with typically developing language skills, and aged between 30-47 months. The 30-minute language samples were audio...
متن کاملlearners’ attitudes toward the effectiveness of mobile-assisted language learning (mall) in vocabulary acquisition in the iranian efl context: the case of word lists, audiobooks and dictionary use
رشد انفجاری تکنولوژی فرصت های آموزشی مهیج و جدیدی را پیش روی فراگیران و آموزش دهندگان گذاشته است. امروزه معلمان برای اینکه در امر آموزش زبان بروز باشند باید روش هایی را اتخاذ نمایند که درآن ها از تکنولوژی جهت کمک در یادگیری زبان دوم و چندم استفاده شده باشد. با در نظر گرفتن تحولاتی که رشته ی آموزش زبان در حال رخ دادن است هم اکنون زمان مناسبی برای ارزشیابی نگرش های موجود نسبت به تکنولوژی های جدید...
15 صفحه اولUsing Factored Word Representation in Neural Network Language Models
Neural network language and translation models have recently shown their great potentials in improving the performance of phrase-based machine translation. At the same time, word representations using different word factors have been translation quality and are part of many state-of-theart machine translation systems. used in many state-of-the-art machine translation systems, in order to suppor...
متن کاملNeural Network Models for Language Acquisition: A Brief Survey
Since the outbreak of connectionist modelling in the mid eighties, several problems in natural language processing have been tackled by employing neural network-based techniques. Neural network's biological plausibility o ers a promising framework in which the computational treatment of language may be linked to other disciplines such as cognitive science and psychology. With this brief survey,...
متن کاملCompressing Neural Language Models by Sparse Word Representations
Neural networks are among the state-ofthe-art techniques for language modeling. Existing neural language models typically map discrete words to distributed, dense vector representations. After information processing of the preceding context words by hidden layers, an output layer estimates the probability of the next word. Such approaches are timeand memory-intensive because of the large number...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Transactions of the Association for Computational Linguistics
سال: 2022
ISSN: ['2307-387X']
DOI: https://doi.org/10.1162/tacl_a_00444